From 3eedad5ee1475f978dc5af17795d92296a27d8d6 Mon Sep 17 00:00:00 2001 From: "kaf24@firebug.cl.cam.ac.uk" Date: Tue, 23 Aug 2005 09:26:11 +0000 Subject: [PATCH] Remove double invocation of vmx_intr_assist when VM_EXIT causes a domain switch. Signed-off-by: Eddie Dong Signed-off-by: Edwin Zhai --- xen/arch/x86/vmx.c | 3 --- xen/arch/x86/vmx_io.c | 9 ++++----- xen/arch/x86/x86_32/entry.S | 1 + xen/arch/x86/x86_64/entry.S | 1 + xen/include/asm-x86/vmx.h | 4 ++-- 5 files changed, 8 insertions(+), 10 deletions(-) diff --git a/xen/arch/x86/vmx.c b/xen/arch/x86/vmx.c index f7c39b8346..d9725938e0 100644 --- a/xen/arch/x86/vmx.c +++ b/xen/arch/x86/vmx.c @@ -1712,9 +1712,6 @@ asmlinkage void vmx_vmexit_handler(struct cpu_user_regs regs) default: __vmx_bug(®s); /* should not happen */ } - - vmx_intr_assist(v); - return; } asmlinkage void load_cr2(void) diff --git a/xen/arch/x86/vmx_io.c b/xen/arch/x86/vmx_io.c index cbc88deb8b..ceee60690c 100644 --- a/xen/arch/x86/vmx_io.c +++ b/xen/arch/x86/vmx_io.c @@ -631,12 +631,14 @@ static inline int irq_masked(unsigned long eflags) return ((eflags & X86_EFLAGS_IF) == 0); } -void vmx_intr_assist(struct vcpu *v) +asmlinkage void vmx_intr_assist(void) { int intr_type = 0; - int highest_vector = find_highest_pending_irq(v, &intr_type); + int highest_vector; unsigned long intr_fields, eflags, interruptibility, cpu_exec_control; + struct vcpu *v = current; + highest_vector = find_highest_pending_irq(v, &intr_type); __vmread(CPU_BASED_VM_EXEC_CONTROL, &cpu_exec_control); if (highest_vector == -1) { @@ -712,9 +714,6 @@ void vmx_do_resume(struct vcpu *d) /* We can't resume the guest if we're waiting on I/O */ ASSERT(!test_bit(ARCH_VMX_IO_WAIT, &d->arch.arch_vmx.flags)); - - /* We always check for interrupts before resuming guest */ - vmx_intr_assist(d); } #endif /* CONFIG_VMX */ diff --git a/xen/arch/x86/x86_32/entry.S b/xen/arch/x86/x86_32/entry.S index a2491612d7..911594b910 100644 --- a/xen/arch/x86/x86_32/entry.S +++ b/xen/arch/x86/x86_32/entry.S @@ -140,6 +140,7 @@ ENTRY(vmx_asm_vmexit_handler) jnz 2f /* vmx_restore_all_guest */ + call vmx_intr_assist call load_cr2 .endif VMX_RESTORE_ALL_NOSEGREGS diff --git a/xen/arch/x86/x86_64/entry.S b/xen/arch/x86/x86_64/entry.S index 0f24b15ca1..3db22487bf 100644 --- a/xen/arch/x86/x86_64/entry.S +++ b/xen/arch/x86/x86_64/entry.S @@ -233,6 +233,7 @@ ENTRY(vmx_asm_vmexit_handler) jnz 2f /* vmx_restore_all_guest */ + call vmx_intr_assist call load_cr2 .endif /* diff --git a/xen/include/asm-x86/vmx.h b/xen/include/asm-x86/vmx.h index 714969eb2e..0d40e51068 100644 --- a/xen/include/asm-x86/vmx.h +++ b/xen/include/asm-x86/vmx.h @@ -31,7 +31,7 @@ extern void vmx_asm_vmexit_handler(struct cpu_user_regs); extern void vmx_asm_do_resume(void); extern void vmx_asm_do_launch(void); -extern void vmx_intr_assist(struct vcpu *d); +extern void vmx_intr_assist(void); extern void arch_vmx_do_launch(struct vcpu *); extern void arch_vmx_do_resume(struct vcpu *); @@ -355,7 +355,7 @@ static inline int __vmxon (u64 addr) } /* Make sure that xen intercepts any FP accesses from current */ -static inline void vmx_stts() +static inline void vmx_stts(void) { unsigned long cr0; -- 2.30.2